This paper focuses on the application of meta-reinforcement learning to the robust design of low-thrust interplanetary trajectories in the presence of multiple uncertainties. A closed-loop control policy is used to optimally steer the spacecraft to a final target state despite the considered perturbations. The control policy is approximated by a deep recurrent neural network, trained by policy-gradient reinforcement learning on a collection of environments featuring mixed sources of uncertainty, namely dynamic uncertainty and control execution errors. The recurrent network is able to build an internal representation of the distribution of environments, thus better adapting the control to the different stochastic scenarios. The results in terms of optimality, constraint handling, and robustness on a fuel-optimal low-thrust transfer between Earth and Mars are compared with those obtained via a traditional reinforcement learning approach based on a feed-forward neural network.

Robust interplanetary trajectory design under multiple uncertainties via meta-reinforcement learning / Federici, Lorenzo; Zavoli, Alessandro. - In: ACTA ASTRONAUTICA. - ISSN 0094-5765. - 214:(2024), pp. 147-158. [10.1016/j.actaastro.2023.10.018]

Robust interplanetary trajectory design under multiple uncertainties via meta-reinforcement learning

Federici, Lorenzo
;
Zavoli, Alessandro
2024

Abstract

This paper focuses on the application of meta-reinforcement learning to the robust design of low-thrust interplanetary trajectories in the presence of multiple uncertainties. A closed-loop control policy is used to optimally steer the spacecraft to a final target state despite the considered perturbations. The control policy is approximated by a deep recurrent neural network, trained by policy-gradient reinforcement learning on a collection of environments featuring mixed sources of uncertainty, namely dynamic uncertainty and control execution errors. The recurrent network is able to build an internal representation of the distribution of environments, thus better adapting the control to the different stochastic scenarios. The results in terms of optimality, constraint handling, and robustness on a fuel-optimal low-thrust transfer between Earth and Mars are compared with those obtained via a traditional reinforcement learning approach based on a feed-forward neural network.
2024
Meta-reinforcement learning; robust trajectory design; closed-loop guidance; recurrent neural network; proximal policy optimization; stochastic optimal control
01 Pubblicazione su rivista::01a Articolo in rivista
Robust interplanetary trajectory design under multiple uncertainties via meta-reinforcement learning / Federici, Lorenzo; Zavoli, Alessandro. - In: ACTA ASTRONAUTICA. - ISSN 0094-5765. - 214:(2024), pp. 147-158. [10.1016/j.actaastro.2023.10.018]
File allegati a questo prodotto
File Dimensione Formato  
Federici_Robust_2024.pdf

accesso aperto

Note: https://www.sciencedirect.com/science/article/pii/S0094576523005258?via=ihub
Tipologia: Documento in Post-print (versione successiva alla peer review e accettata per la pubblicazione)
Licenza: Creative commons
Dimensione 1.25 MB
Formato Adobe PDF
1.25 MB Adobe PDF
Federici_Robust_2024.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 2.53 MB
Formato Adobe PDF
2.53 MB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1713983
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 11
  • ???jsp.display-item.citation.isi??? 9
social impact